Skip to content

Copilot skill for Fabric Lakehouse#739

Closed
tedvilutis wants to merge 5 commits intogithub:mainfrom
tedvilutis:main
Closed

Copilot skill for Fabric Lakehouse#739
tedvilutis wants to merge 5 commits intogithub:mainfrom
tedvilutis:main

Conversation

@tedvilutis
Copy link
Copy Markdown
Contributor

Pull Request Checklist

  • I have read and followed the CONTRIBUTING.md guidelines.
  • My contribution adds a new instruction, prompt, agent, or skill file in the correct directory.
  • The file follows the required naming convention.
  • The content is clearly structured and follows the example format.
  • I have tested my instructions, prompt, agent, or skill with GitHub Copilot.
  • I have run npm start and verified that README.md is up to date.

Description

This is a skill explaining to agents what Fabric Lakehouse is. I'm a product manager of Microsoft Fabric team owning Lakehouse experience. When using Copilot to generate PRD or Code that is related to Lakehouse without this skill you get some discrepancies with the actual product. This skill will let agents to be more accurate when working on fabric Lakehouse.


Type of Contribution

  • New instruction file.
  • New prompt file.
  • New agent file.
  • New plugin.
  • New skill file.
  • Update to existing instruction, prompt, agent, plugin, or skill.
  • Other (please specify):

Additional Notes

I'll continue maintaining and updating this skill as Fabric Lakehouse evolves.


By submitting this pull request, I confirm that my contribution abides by the Code of Conduct and will be licensed under the MIT License.

tedvilutis and others added 3 commits February 16, 2026 12:23
This skill explain Copilot agents what Fabric Lakehouse is and what features and capabilities it has.
Copilot AI review requested due to automatic review settings February 16, 2026 23:48
Copy link
Copy Markdown
Contributor

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull request overview

This pull request adds a new skill for Microsoft Fabric Lakehouse to help AI agents provide accurate information when working with Fabric Lakehouse-related tasks. The skill provides comprehensive documentation about Lakehouse concepts, architecture, security, performance optimization, and code examples.

Changes:

  • Added a new fabric-lakehouse skill with SKILL.md and reference documentation
  • Included PySpark code examples for common Lakehouse operations (reading/writing data, Delta operations, optimization)
  • Provided Data Factory integration patterns for ETL/ELT orchestration

Reviewed changes

Copilot reviewed 3 out of 3 changed files in this pull request and generated 14 comments.

File Description
skills/fabric-lakehouse/SKILL.md Main skill definition with frontmatter, core concepts, security, shortcuts, optimization, and lineage information
skills/fabric-lakehouse/references/pyspark.md PySpark code examples for Spark configuration, data reading/writing, Delta operations, schema definition, and optimization patterns
skills/fabric-lakehouse/references/getdata.md Data Factory integration documentation including connectors, pipeline activities, and orchestration patterns

Comment thread skills/fabric-lakehouse/SKILL.md Outdated
### Tabular data in a Lakehouse

Tabular data in a form of tables are stored under "Tables" folder. Main format for tables in Lakehouse is Delta. Lakehouse can store tabular data in other formats like CSV or Parquet, these formats only available for Spark querying.
Tables can be internal, when data is stored under "Tables" folder" or external, when only reference to a table is stored under "Tables" folder but the data itself is stored in a referenced location. Referecing tables are done through Shortcuts, which can be internal, pointing to other location in Fabric, or external pointing to data stored outside of Fabric.
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Spelling error: "Referecing" should be "Referencing".

Suggested change
Tables can be internal, when data is stored under "Tables" folder" or external, when only reference to a table is stored under "Tables" folder but the data itself is stored in a referenced location. Referecing tables are done through Shortcuts, which can be internal, pointing to other location in Fabric, or external pointing to data stored outside of Fabric.
Tables can be internal, when data is stored under "Tables" folder" or external, when only reference to a table is stored under "Tables" folder but the data itself is stored in a referenced location. Referencing tables are done through Shortcuts, which can be internal, pointing to other location in Fabric, or external pointing to data stored outside of Fabric.

Copilot uses AI. Check for mistakes.
Comment thread skills/fabric-lakehouse/SKILL.md Outdated

### Schemas for tables in a Lakehouse

When creating a lakehouse user can choose to enable schemas. Schemas are used to organize Lakehouse tables. Schemas are implemented as folders under "Tables" folder and store tables inside of those folders. Default schema is "dbo" and it can't be deleted or renamed. All other schemas are optional and can be created, renamed, or deleted. User can reference schema located in other lakehouse using Schema Shortcut that way referincing all tables with one shortcut that are at the destination schema.
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Spelling error: "referincing" should be "referencing".

Suggested change
When creating a lakehouse user can choose to enable schemas. Schemas are used to organize Lakehouse tables. Schemas are implemented as folders under "Tables" folder and store tables inside of those folders. Default schema is "dbo" and it can't be deleted or renamed. All other schemas are optional and can be created, renamed, or deleted. User can reference schema located in other lakehouse using Schema Shortcut that way referincing all tables with one shortcut that are at the destination schema.
When creating a lakehouse user can choose to enable schemas. Schemas are used to organize Lakehouse tables. Schemas are implemented as folders under "Tables" folder and store tables inside of those folders. Default schema is "dbo" and it can't be deleted or renamed. All other schemas are optional and can be created, renamed, or deleted. User can reference schema located in other lakehouse using Schema Shortcut that way referencing all tables with one shortcut that are at the destination schema.

Copilot uses AI. Check for mistakes.
Comment thread skills/fabric-lakehouse/SKILL.md Outdated

### Fabric Materialized Views

Set of pre-computed tables that are automatically updated based on schedule. They provide fast query performance for complex aggregations and joins. Materialized views are defined using PySpark or Spark SQL stored in asociated Notebook.
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Spelling error: "asociated" should be "associated".

Suggested change
Set of pre-computed tables that are automatically updated based on schedule. They provide fast query performance for complex aggregations and joins. Materialized views are defined using PySpark or Spark SQL stored in asociated Notebook.
Set of pre-computed tables that are automatically updated based on schedule. They provide fast query performance for complex aggregations and joins. Materialized views are defined using PySpark or Spark SQL stored in associated Notebook.

Copilot uses AI. Check for mistakes.
Comment thread skills/fabric-lakehouse/SKILL.md Outdated

### Data access or OneLake Security

For data access use OneLake security model, which is based on Azure Active Directory (AAD) and role-based access control (RBAC). Lakehouse data is stored in OneLake, so access to data is controlled through OneLake permissions. In adition to object-level permissions, Lakehouse also supports column-level and row-level security for tables, allowing fine-grained control over who can see specific columns or rows in a table.
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Spelling error: "adition" should be "addition".

Suggested change
For data access use OneLake security model, which is based on Azure Active Directory (AAD) and role-based access control (RBAC). Lakehouse data is stored in OneLake, so access to data is controlled through OneLake permissions. In adition to object-level permissions, Lakehouse also supports column-level and row-level security for tables, allowing fine-grained control over who can see specific columns or rows in a table.
For data access use OneLake security model, which is based on Azure Active Directory (AAD) and role-based access control (RBAC). Lakehouse data is stored in OneLake, so access to data is controlled through OneLake permissions. In addition to object-level permissions, Lakehouse also supports column-level and row-level security for tables, allowing fine-grained control over who can see specific columns or rows in a table.

Copilot uses AI. Check for mistakes.

### V-Order Optimization

For faster data read with semantic model enable V-Order optimization on Delta tables.This presorts data in a way that improves query performance for common access patterns.
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Missing space after period. Should be "tables. This" instead of "tables.This".

Suggested change
For faster data read with semantic model enable V-Order optimization on Delta tables.This presorts data in a way that improves query performance for common access patterns.
For faster data read with semantic model enable V-Order optimization on Delta tables. This presorts data in a way that improves query performance for common access patterns.

Copilot uses AI. Check for mistakes.
Comment thread skills/fabric-lakehouse/SKILL.md Outdated

### Files in a Lakehouse

Files are stored uner "Files" folder. Users can create folders and subfolders to organize their files. Any file format can be stored in Lakehosue.
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Spelling error: "uner" should be "under".

Suggested change
Files are stored uner "Files" folder. Users can create folders and subfolders to organize their files. Any file format can be stored in Lakehosue.
Files are stored under "Files" folder. Users can create folders and subfolders to organize their files. Any file format can be stored in Lakehosue.

Copilot uses AI. Check for mistakes.
Comment thread skills/fabric-lakehouse/SKILL.md Outdated

### Files in a Lakehouse

Files are stored uner "Files" folder. Users can create folders and subfolders to organize their files. Any file format can be stored in Lakehosue.
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Spelling error: "Lakehosue" should be "Lakehouse".

Copilot uses AI. Check for mistakes.
Comment thread skills/fabric-lakehouse/SKILL.md Outdated
### Tabular data in a Lakehouse

Tabular data in a form of tables are stored under "Tables" folder. Main format for tables in Lakehouse is Delta. Lakehouse can store tabular data in other formats like CSV or Parquet, these formats only available for Spark querying.
Tables can be internal, when data is stored under "Tables" folder" or external, when only reference to a table is stored under "Tables" folder but the data itself is stored in a referenced location. Referecing tables are done through Shortcuts, which can be internal, pointing to other location in Fabric, or external pointing to data stored outside of Fabric.
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Grammar issue: Extra quotation mark in the phrase. Should be 'under "Tables" folder' instead of 'under "Tables" folder"'. The closing quote after "folder" should be removed.

Suggested change
Tables can be internal, when data is stored under "Tables" folder" or external, when only reference to a table is stored under "Tables" folder but the data itself is stored in a referenced location. Referecing tables are done through Shortcuts, which can be internal, pointing to other location in Fabric, or external pointing to data stored outside of Fabric.
Tables can be internal, when data is stored under "Tables" folder or external, when only reference to a table is stored under "Tables" folder but the data itself is stored in a referenced location. Referecing tables are done through Shortcuts, which can be internal, pointing to other location in Fabric, or external pointing to data stored outside of Fabric.

Copilot uses AI. Check for mistakes.
Comment thread skills/fabric-lakehouse/SKILL.md Outdated

### Data access or OneLake Security

For data access use OneLake security model, which is based on Azure Active Directory (AAD) and role-based access control (RBAC). Lakehouse data is stored in OneLake, so access to data is controlled through OneLake permissions. In adition to object-level permissions, Lakehouse also supports column-level and row-level security for tables, allowing fine-grained control over who can see specific columns or rows in a table.
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Outdated terminology: "Azure Active Directory (AAD)" has been rebranded to "Microsoft Entra ID". Consider updating the terminology to reflect the current product name.

Suggested change
For data access use OneLake security model, which is based on Azure Active Directory (AAD) and role-based access control (RBAC). Lakehouse data is stored in OneLake, so access to data is controlled through OneLake permissions. In adition to object-level permissions, Lakehouse also supports column-level and row-level security for tables, allowing fine-grained control over who can see specific columns or rows in a table.
For data access use OneLake security model, which is based on Microsoft Entra ID (formerly Azure Active Directory) and role-based access control (RBAC). Lakehouse data is stored in OneLake, so access to data is controlled through OneLake permissions. In adition to object-level permissions, Lakehouse also supports column-level and row-level security for tables, allowing fine-grained control over who can see specific columns or rows in a table.

Copilot uses AI. Check for mistakes.
Comment thread skills/fabric-lakehouse/SKILL.md Outdated

### Key Components

- **Delta Tables** Managed tables with ACID compliance and schema enforcement
Copy link

Copilot AI Feb 16, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Inconsistent list formatting: "Delta Tables" on line 27 is missing a separator (colon, dash, or em dash) between the term and its description. For consistency with other list items like "Files" (line 28) and "SQL Endpoint" (line 29), consider adding a separator such as a colon or dash after "Delta Tables".

Suggested change
- **Delta Tables** Managed tables with ACID compliance and schema enforcement
- **Delta Tables**: Managed tables with ACID compliance and schema enforcement

Copilot uses AI. Check for mistakes.
@tedvilutis tedvilutis closed this by deleting the head repository Feb 17, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants